2026-03-17 05:27:51,448 INFO mapreduce.MiniHadoopClusterManager: Updated 0 configuration settings from command line. 2026-03-17 05:27:51,490 INFO hdfs.MiniDFSCluster: starting cluster: numNameNodes=1, numDataNodes=1 2026-03-17 05:27:51,799 INFO namenode.NameNode: Formatting using clusterid: testClusterID 2026-03-17 05:27:51,810 INFO namenode.FSEditLog: Edit logging is async:true 2026-03-17 05:27:51,826 INFO namenode.FSNamesystem: KeyProvider: null 2026-03-17 05:27:51,827 INFO namenode.FSNamesystem: fsLock is fair: true 2026-03-17 05:27:51,827 INFO namenode.FSNamesystem: Detailed lock hold time metrics enabled: false 2026-03-17 05:27:51,848 INFO namenode.FSNamesystem: fsOwner = clickhouse (auth:SIMPLE) 2026-03-17 05:27:51,848 INFO namenode.FSNamesystem: supergroup = supergroup 2026-03-17 05:27:51,848 INFO namenode.FSNamesystem: isPermissionEnabled = true 2026-03-17 05:27:51,848 INFO namenode.FSNamesystem: isStoragePolicyEnabled = true 2026-03-17 05:27:51,848 INFO namenode.FSNamesystem: HA Enabled: false 2026-03-17 05:27:51,874 INFO common.Util: dfs.datanode.fileio.profiling.sampling.percentage set to 0. Disabling file IO profiling 2026-03-17 05:27:51,876 INFO Configuration.deprecation: hadoop.configured.node.mapping is deprecated. Instead, use net.topology.configured.node.mapping 2026-03-17 05:27:51,876 INFO blockmanagement.DatanodeManager: dfs.block.invalidate.limit: configured=1000, counted=60, effected=1000 2026-03-17 05:27:51,876 INFO blockmanagement.DatanodeManager: dfs.namenode.datanode.registration.ip-hostname-check=true 2026-03-17 05:27:51,878 INFO blockmanagement.BlockManager: dfs.namenode.startup.delay.block.deletion.sec is set to 000:00:00:00.000 2026-03-17 05:27:51,878 INFO blockmanagement.BlockManager: The block deletion will start around 2026 Mar 17 05:27:51 2026-03-17 05:27:51,879 INFO util.GSet: Computing capacity for map BlocksMap 2026-03-17 05:27:51,879 INFO util.GSet: VM type = 64-bit 2026-03-17 05:27:51,880 INFO util.GSet: 2.0% max memory 7.7 GB = 156.7 MB 2026-03-17 05:27:51,880 INFO util.GSet: capacity = 2^24 = 16777216 entries 2026-03-17 05:27:51,914 INFO blockmanagement.BlockManager: Storage policy satisfier is disabled 2026-03-17 05:27:51,914 INFO blockmanagement.BlockManager: dfs.block.access.token.enable = false 2026-03-17 05:27:51,917 INFO blockmanagement.BlockManagerSafeMode: dfs.namenode.safemode.threshold-pct = 0.999 2026-03-17 05:27:51,917 INFO blockmanagement.BlockManagerSafeMode: dfs.namenode.safemode.min.datanodes = 0 2026-03-17 05:27:51,917 INFO blockmanagement.BlockManagerSafeMode: dfs.namenode.safemode.extension = 0 2026-03-17 05:27:51,917 INFO blockmanagement.BlockManager: defaultReplication = 1 2026-03-17 05:27:51,918 INFO blockmanagement.BlockManager: maxReplication = 512 2026-03-17 05:27:51,918 INFO blockmanagement.BlockManager: minReplication = 1 2026-03-17 05:27:51,918 INFO blockmanagement.BlockManager: maxReplicationStreams = 2 2026-03-17 05:27:51,918 INFO blockmanagement.BlockManager: redundancyRecheckInterval = 3000ms 2026-03-17 05:27:51,918 INFO blockmanagement.BlockManager: encryptDataTransfer = false 2026-03-17 05:27:51,918 INFO blockmanagement.BlockManager: maxNumBlocksToLog = 1000 2026-03-17 05:27:51,930 INFO namenode.FSDirectory: GLOBAL serial map: bits=29 maxEntries=536870911 2026-03-17 05:27:51,930 INFO namenode.FSDirectory: USER serial map: bits=24 maxEntries=16777215 2026-03-17 05:27:51,930 INFO namenode.FSDirectory: GROUP serial map: bits=24 maxEntries=16777215 2026-03-17 05:27:51,930 INFO namenode.FSDirectory: XATTR serial map: bits=24 maxEntries=16777215 2026-03-17 05:27:51,936 INFO util.GSet: Computing capacity for map INodeMap 2026-03-17 05:27:51,936 INFO util.GSet: VM type = 64-bit 2026-03-17 05:27:51,937 INFO util.GSet: 1.0% max memory 7.7 GB = 78.3 MB 2026-03-17 05:27:51,937 INFO util.GSet: capacity = 2^23 = 8388608 entries 2026-03-17 05:27:51,951 INFO namenode.FSDirectory: ACLs enabled? true 2026-03-17 05:27:51,951 INFO namenode.FSDirectory: POSIX ACL inheritance enabled? true 2026-03-17 05:27:51,951 INFO namenode.FSDirectory: XAttrs enabled? true 2026-03-17 05:27:51,952 INFO namenode.NameNode: Caching file names occurring more than 10 times 2026-03-17 05:27:51,954 INFO snapshot.SnapshotManager: Loaded config captureOpenFiles: false, skipCaptureAccessTimeOnlyChange: false, snapshotDiffAllowSnapRootDescendant: true, maxSnapshotLimit: 65536 2026-03-17 05:27:51,956 INFO snapshot.SnapshotManager: SkipList is disabled 2026-03-17 05:27:51,958 INFO util.GSet: Computing capacity for map cachedBlocks 2026-03-17 05:27:51,958 INFO util.GSet: VM type = 64-bit 2026-03-17 05:27:51,959 INFO util.GSet: 0.25% max memory 7.7 GB = 19.6 MB 2026-03-17 05:27:51,959 INFO util.GSet: capacity = 2^21 = 2097152 entries 2026-03-17 05:27:51,968 INFO metrics.TopMetrics: NNTop conf: dfs.namenode.top.window.num.buckets = 10 2026-03-17 05:27:51,968 INFO metrics.TopMetrics: NNTop conf: dfs.namenode.top.num.users = 10 2026-03-17 05:27:51,968 INFO metrics.TopMetrics: NNTop conf: dfs.namenode.top.windows.minutes = 1,5,25 2026-03-17 05:27:51,970 INFO namenode.FSNamesystem: Retry cache on namenode is enabled 2026-03-17 05:27:51,970 INFO namenode.FSNamesystem: Retry cache will use 0.03 of total heap and retry cache entry expiry time is 600000 millis 2026-03-17 05:27:51,971 INFO util.GSet: Computing capacity for map NameNodeRetryCache 2026-03-17 05:27:51,971 INFO util.GSet: VM type = 64-bit 2026-03-17 05:27:51,971 INFO util.GSet: 0.029999999329447746% max memory 7.7 GB = 2.4 MB 2026-03-17 05:27:51,971 INFO util.GSet: capacity = 2^18 = 262144 entries 2026-03-17 05:27:51,988 INFO namenode.FSImage: Allocated new BlockPoolId: BP-1367658773-172.17.0.2-1773761271981 2026-03-17 05:27:51,997 INFO common.Storage: Storage directory /hadoop-3.3.1/target/test/data/dfs/name-0-1 has been successfully formatted. 2026-03-17 05:27:52,001 INFO common.Storage: Storage directory /hadoop-3.3.1/target/test/data/dfs/name-0-2 has been successfully formatted. 2026-03-17 05:27:52,052 INFO namenode.FSImageFormatProtobuf: Saving image file /hadoop-3.3.1/target/test/data/dfs/name-0-1/current/fsimage.ckpt_0000000000000000000 using no compression 2026-03-17 05:27:52,052 INFO namenode.FSImageFormatProtobuf: Saving image file /hadoop-3.3.1/target/test/data/dfs/name-0-2/current/fsimage.ckpt_0000000000000000000 using no compression 2026-03-17 05:27:52,136 INFO namenode.FSImageFormatProtobuf: Image file /hadoop-3.3.1/target/test/data/dfs/name-0-1/current/fsimage.ckpt_0000000000000000000 of size 405 bytes saved in 0 seconds . 2026-03-17 05:27:52,136 INFO namenode.FSImageFormatProtobuf: Image file /hadoop-3.3.1/target/test/data/dfs/name-0-2/current/fsimage.ckpt_0000000000000000000 of size 405 bytes saved in 0 seconds . 2026-03-17 05:27:52,148 INFO namenode.NNStorageRetentionManager: Going to retain 1 images with txid >= 0 2026-03-17 05:27:52,185 INFO namenode.FSNamesystem: Stopping services started for active state 2026-03-17 05:27:52,186 INFO namenode.FSNamesystem: Stopping services started for standby state 2026-03-17 05:27:52,187 INFO namenode.NameNode: createNameNode [] 2026-03-17 05:27:52,222 INFO impl.MetricsConfig: Loaded properties from hadoop-metrics2.properties 2026-03-17 05:27:52,276 INFO impl.MetricsSystemImpl: Scheduled Metric snapshot period at 10 second(s). 2026-03-17 05:27:52,276 INFO impl.MetricsSystemImpl: NameNode metrics system started 2026-03-17 05:27:52,289 INFO namenode.NameNodeUtils: fs.defaultFS is hdfs://127.0.0.1:12222 2026-03-17 05:27:52,289 INFO namenode.NameNode: Clients should use 127.0.0.1:12222 to access this namenode/service. 2026-03-17 05:27:52,329 INFO util.JvmPauseMonitor: Starting JVM pause monitor 2026-03-17 05:27:52,334 INFO hdfs.DFSUtil: Filter initializers set : org.apache.hadoop.http.lib.StaticUserWebFilter,org.apache.hadoop.hdfs.web.AuthFilterInitializer 2026-03-17 05:27:52,337 INFO hdfs.DFSUtil: Starting Web-server for hdfs at: http://localhost:0 2026-03-17 05:27:52,344 INFO util.log: Logging initialized @1287ms to org.eclipse.jetty.util.log.Slf4jLog 2026-03-17 05:27:52,389 INFO server.AuthenticationFilter: Unable to initialize FileSignerSecretProvider, falling back to use random secrets. 2026-03-17 05:27:52,391 INFO http.HttpRequestLog: Http request log for http.requests.namenode is not defined 2026-03-17 05:27:52,395 INFO http.HttpServer2: Added global filter 'safety' (class=org.apache.hadoop.http.HttpServer2$QuotingInputFilter) 2026-03-17 05:27:52,396 INFO http.HttpServer2: Added filter static_user_filter (class=org.apache.hadoop.http.lib.StaticUserWebFilter$StaticUserFilter) to context hdfs 2026-03-17 05:27:52,396 INFO http.HttpServer2: Added filter static_user_filter (class=org.apache.hadoop.http.lib.StaticUserWebFilter$StaticUserFilter) to context logs 2026-03-17 05:27:52,396 INFO http.HttpServer2: Added filter static_user_filter (class=org.apache.hadoop.http.lib.StaticUserWebFilter$StaticUserFilter) to context static 2026-03-17 05:27:52,397 INFO http.HttpServer2: Added filter AuthFilter (class=org.apache.hadoop.hdfs.web.AuthFilter) to context hdfs 2026-03-17 05:27:52,397 INFO http.HttpServer2: Added filter AuthFilter (class=org.apache.hadoop.hdfs.web.AuthFilter) to context logs 2026-03-17 05:27:52,397 INFO http.HttpServer2: Added filter AuthFilter (class=org.apache.hadoop.hdfs.web.AuthFilter) to context static 2026-03-17 05:27:52,413 INFO http.HttpServer2: addJerseyResourcePackage: packageName=org.apache.hadoop.hdfs.server.namenode.web.resources;org.apache.hadoop.hdfs.web.resources, pathSpec=/webhdfs/v1/* 2026-03-17 05:27:52,418 INFO http.HttpServer2: Jetty bound to port 46257 2026-03-17 05:27:52,419 INFO server.Server: jetty-9.4.40.v20210413; built: 2021-04-13T20:42:42.668Z; git: b881a572662e1943a14ae12e7e1207989f218b74; jvm 11.0.25+9-post-Ubuntu-1ubuntu122.04 2026-03-17 05:27:52,434 INFO server.session: DefaultSessionIdManager workerName=node0 2026-03-17 05:27:52,434 INFO server.session: No SessionScavenger set, using defaults 2026-03-17 05:27:52,435 INFO server.session: node0 Scavenging every 660000ms 2026-03-17 05:27:52,443 INFO server.AuthenticationFilter: Unable to initialize FileSignerSecretProvider, falling back to use random secrets. 2026-03-17 05:27:52,444 INFO handler.ContextHandler: Started o.e.j.s.ServletContextHandler@405325cf{logs,/logs,file:///hadoop-3.3.1/logs,AVAILABLE} 2026-03-17 05:27:52,444 INFO handler.ContextHandler: Started o.e.j.s.ServletContextHandler@5d52e3ef{static,/static,file:///hadoop-3.3.1/share/hadoop/hdfs/webapps/static/,AVAILABLE} 2026-03-17 05:27:52,492 INFO handler.ContextHandler: Started o.e.j.w.WebAppContext@62ddd21b{hdfs,/,file:///hadoop-3.3.1/share/hadoop/hdfs/webapps/hdfs/,AVAILABLE}{file:/hadoop-3.3.1/share/hadoop/hdfs/webapps/hdfs} 2026-03-17 05:27:52,503 INFO server.AbstractConnector: Started ServerConnector@72889280{HTTP/1.1, (http/1.1)}{localhost:46257} 2026-03-17 05:27:52,503 INFO server.Server: Started @1446ms 2026-03-17 05:27:52,506 INFO namenode.FSEditLog: Edit logging is async:true 2026-03-17 05:27:52,514 INFO namenode.FSNamesystem: KeyProvider: null 2026-03-17 05:27:52,514 INFO namenode.FSNamesystem: fsLock is fair: true 2026-03-17 05:27:52,514 INFO namenode.FSNamesystem: Detailed lock hold time metrics enabled: false 2026-03-17 05:27:52,514 INFO namenode.FSNamesystem: fsOwner = clickhouse (auth:SIMPLE) 2026-03-17 05:27:52,514 INFO namenode.FSNamesystem: supergroup = supergroup 2026-03-17 05:27:52,514 INFO namenode.FSNamesystem: isPermissionEnabled = true 2026-03-17 05:27:52,514 INFO namenode.FSNamesystem: isStoragePolicyEnabled = true 2026-03-17 05:27:52,514 INFO namenode.FSNamesystem: HA Enabled: false 2026-03-17 05:27:52,514 INFO common.Util: dfs.datanode.fileio.profiling.sampling.percentage set to 0. Disabling file IO profiling 2026-03-17 05:27:52,514 INFO blockmanagement.DatanodeManager: dfs.block.invalidate.limit: configured=1000, counted=60, effected=1000 2026-03-17 05:27:52,514 INFO blockmanagement.DatanodeManager: dfs.namenode.datanode.registration.ip-hostname-check=true 2026-03-17 05:27:52,515 INFO blockmanagement.BlockManager: dfs.namenode.startup.delay.block.deletion.sec is set to 000:00:00:00.000 2026-03-17 05:27:52,515 INFO blockmanagement.BlockManager: The block deletion will start around 2026 Mar 17 05:27:52 2026-03-17 05:27:52,515 INFO util.GSet: Computing capacity for map BlocksMap 2026-03-17 05:27:52,515 INFO util.GSet: VM type = 64-bit 2026-03-17 05:27:52,515 INFO util.GSet: 2.0% max memory 7.7 GB = 156.7 MB 2026-03-17 05:27:52,515 INFO util.GSet: capacity = 2^24 = 16777216 entries 2026-03-17 05:27:52,516 INFO blockmanagement.BlockManager: Storage policy satisfier is disabled 2026-03-17 05:27:52,516 INFO blockmanagement.BlockManager: dfs.block.access.token.enable = false 2026-03-17 05:27:52,517 INFO blockmanagement.BlockManagerSafeMode: dfs.namenode.safemode.threshold-pct = 0.999 2026-03-17 05:27:52,517 INFO blockmanagement.BlockManagerSafeMode: dfs.namenode.safemode.min.datanodes = 0 2026-03-17 05:27:52,517 INFO blockmanagement.BlockManagerSafeMode: dfs.namenode.safemode.extension = 0 2026-03-17 05:27:52,517 INFO blockmanagement.BlockManager: defaultReplication = 1 2026-03-17 05:27:52,517 INFO blockmanagement.BlockManager: maxReplication = 512 2026-03-17 05:27:52,517 INFO blockmanagement.BlockManager: minReplication = 1 2026-03-17 05:27:52,517 INFO blockmanagement.BlockManager: maxReplicationStreams = 2 2026-03-17 05:27:52,517 INFO blockmanagement.BlockManager: redundancyRecheckInterval = 3000ms 2026-03-17 05:27:52,517 INFO blockmanagement.BlockManager: encryptDataTransfer = false 2026-03-17 05:27:52,517 INFO blockmanagement.BlockManager: maxNumBlocksToLog = 1000 2026-03-17 05:27:52,517 INFO util.GSet: Computing capacity for map INodeMap 2026-03-17 05:27:52,517 INFO util.GSet: VM type = 64-bit 2026-03-17 05:27:52,517 INFO util.GSet: 1.0% max memory 7.7 GB = 78.3 MB 2026-03-17 05:27:52,517 INFO util.GSet: capacity = 2^23 = 8388608 entries 2026-03-17 05:27:52,533 INFO namenode.FSDirectory: ACLs enabled? true 2026-03-17 05:27:52,534 INFO namenode.FSDirectory: POSIX ACL inheritance enabled? true 2026-03-17 05:27:52,534 INFO namenode.FSDirectory: XAttrs enabled? true 2026-03-17 05:27:52,534 INFO namenode.NameNode: Caching file names occurring more than 10 times 2026-03-17 05:27:52,534 INFO snapshot.SnapshotManager: Loaded config captureOpenFiles: false, skipCaptureAccessTimeOnlyChange: false, snapshotDiffAllowSnapRootDescendant: true, maxSnapshotLimit: 65536 2026-03-17 05:27:52,534 INFO snapshot.SnapshotManager: SkipList is disabled 2026-03-17 05:27:52,534 INFO util.GSet: Computing capacity for map cachedBlocks 2026-03-17 05:27:52,534 INFO util.GSet: VM type = 64-bit 2026-03-17 05:27:52,534 INFO util.GSet: 0.25% max memory 7.7 GB = 19.6 MB 2026-03-17 05:27:52,534 INFO util.GSet: capacity = 2^21 = 2097152 entries 2026-03-17 05:27:52,538 INFO metrics.TopMetrics: NNTop conf: dfs.namenode.top.window.num.buckets = 10 2026-03-17 05:27:52,538 INFO metrics.TopMetrics: NNTop conf: dfs.namenode.top.num.users = 10 2026-03-17 05:27:52,538 INFO metrics.TopMetrics: NNTop conf: dfs.namenode.top.windows.minutes = 1,5,25 2026-03-17 05:27:52,538 INFO namenode.FSNamesystem: Retry cache on namenode is enabled 2026-03-17 05:27:52,538 INFO namenode.FSNamesystem: Retry cache will use 0.03 of total heap and retry cache entry expiry time is 600000 millis 2026-03-17 05:27:52,538 INFO util.GSet: Computing capacity for map NameNodeRetryCache 2026-03-17 05:27:52,538 INFO util.GSet: VM type = 64-bit 2026-03-17 05:27:52,538 INFO util.GSet: 0.029999999329447746% max memory 7.7 GB = 2.4 MB 2026-03-17 05:27:52,538 INFO util.GSet: capacity = 2^18 = 262144 entries 2026-03-17 05:27:52,545 INFO common.Storage: Lock on /hadoop-3.3.1/target/test/data/dfs/name-0-1/in_use.lock acquired by nodename 413@8de11e8011ac 2026-03-17 05:27:52,547 INFO common.Storage: Lock on /hadoop-3.3.1/target/test/data/dfs/name-0-2/in_use.lock acquired by nodename 413@8de11e8011ac 2026-03-17 05:27:52,548 INFO namenode.FileJournalManager: Recovering unfinalized segments in /hadoop-3.3.1/target/test/data/dfs/name-0-1/current 2026-03-17 05:27:52,548 INFO namenode.FileJournalManager: Recovering unfinalized segments in /hadoop-3.3.1/target/test/data/dfs/name-0-2/current 2026-03-17 05:27:52,548 INFO namenode.FSImage: No edit log streams selected. 2026-03-17 05:27:52,549 INFO namenode.FSImage: Planning to load image: FSImageFile(file=/hadoop-3.3.1/target/test/data/dfs/name-0-1/current/fsimage_0000000000000000000, cpktTxId=0000000000000000000) 2026-03-17 05:27:52,561 INFO namenode.FSImageFormatPBINode: Loading 1 INodes. 2026-03-17 05:27:52,561 INFO namenode.FSImageFormatPBINode: Successfully loaded 1 inodes 2026-03-17 05:27:52,564 INFO namenode.FSImageFormatPBINode: Completed update blocks map and name cache, total waiting duration 0ms. 2026-03-17 05:27:52,565 INFO namenode.FSImageFormatProtobuf: Loaded FSImage in 0 seconds. 2026-03-17 05:27:52,565 INFO namenode.FSImage: Loaded image for txid 0 from /hadoop-3.3.1/target/test/data/dfs/name-0-1/current/fsimage_0000000000000000000 2026-03-17 05:27:52,567 INFO namenode.FSNamesystem: Need to save fs image? false (staleImage=false, haEnabled=false, isRollingUpgrade=false) 2026-03-17 05:27:52,568 INFO namenode.FSEditLog: Starting log segment at 1 2026-03-17 05:27:52,581 INFO namenode.NameCache: initialized with 0 entries 0 lookups 2026-03-17 05:27:52,581 INFO namenode.FSNamesystem: Finished loading FSImage in 41 msecs 2026-03-17 05:27:52,654 INFO namenode.NameNode: RPC server is binding to localhost:12222 2026-03-17 05:27:52,654 INFO namenode.NameNode: Enable NameNode state context:false 2026-03-17 05:27:52,659 INFO ipc.CallQueueManager: Using callQueue: class java.util.concurrent.LinkedBlockingQueue, queueCapacity: 1000, scheduler: class org.apache.hadoop.ipc.DefaultRpcScheduler, ipcBackoff: false. 2026-03-17 05:27:52,664 INFO ipc.Server: Starting Socket Reader #1 for port 12222 2026-03-17 05:27:52,764 INFO namenode.FSNamesystem: Registered FSNamesystemState, ReplicatedBlocksState and ECBlockGroupsState MBeans. 2026-03-17 05:27:52,781 INFO namenode.LeaseManager: Number of blocks under construction: 0 2026-03-17 05:27:52,786 INFO blockmanagement.DatanodeAdminDefaultMonitor: Initialized the Default Decommission and Maintenance monitor 2026-03-17 05:27:52,788 INFO blockmanagement.BlockManager: initializing replication queues 2026-03-17 05:27:52,788 INFO hdfs.StateChange: STATE* Leaving safe mode after 0 secs 2026-03-17 05:27:52,788 INFO hdfs.StateChange: STATE* Network topology has 0 racks and 0 datanodes 2026-03-17 05:27:52,788 INFO hdfs.StateChange: STATE* UnderReplicatedBlocks has 0 blocks 2026-03-17 05:27:52,799 INFO blockmanagement.BlockManager: Total number of blocks = 0 2026-03-17 05:27:52,800 INFO blockmanagement.BlockManager: Number of invalid blocks = 0 2026-03-17 05:27:52,800 INFO blockmanagement.BlockManager: Number of under-replicated blocks = 0 2026-03-17 05:27:52,800 INFO blockmanagement.BlockManager: Number of over-replicated blocks = 0 2026-03-17 05:27:52,800 INFO blockmanagement.BlockManager: Number of blocks being written = 0 2026-03-17 05:27:52,800 INFO hdfs.StateChange: STATE* Replication Queue initialization scan for invalid, over- and under-replicated blocks completed in 12 msec 2026-03-17 05:27:52,808 INFO ipc.Server: IPC Server Responder: starting 2026-03-17 05:27:52,808 INFO ipc.Server: IPC Server listener on 12222: starting 2026-03-17 05:27:52,811 INFO namenode.NameNode: NameNode RPC up at: localhost/127.0.0.1:12222 2026-03-17 05:27:52,813 INFO namenode.FSNamesystem: Starting services required for active state 2026-03-17 05:27:52,813 INFO namenode.FSDirectory: Initializing quota with 12 thread(s) 2026-03-17 05:27:52,817 INFO namenode.FSDirectory: Quota initialization completed in 4 milliseconds name space=1 storage space=0 storage types=RAM_DISK=0, SSD=0, DISK=0, ARCHIVE=0, PROVIDED=0 2026-03-17 05:27:52,835 INFO blockmanagement.CacheReplicationMonitor: Starting CacheReplicationMonitor with interval 30000 milliseconds 2026-03-17 05:27:52,842 INFO hdfs.MiniDFSCluster: Starting DataNode 0 with dfs.datanode.data.dir: [DISK]file:/hadoop-3.3.1/target/test/data/dfs/data/data1,[DISK]file:/hadoop-3.3.1/target/test/data/dfs/data/data2 2026-03-17 05:27:52,870 INFO checker.ThrottledAsyncChecker: Scheduling a check for [DISK]file:/hadoop-3.3.1/target/test/data/dfs/data/data1 2026-03-17 05:27:52,876 INFO checker.ThrottledAsyncChecker: Scheduling a check for [DISK]file:/hadoop-3.3.1/target/test/data/dfs/data/data2 2026-03-17 05:27:52,891 INFO impl.MetricsSystemImpl: DataNode metrics system started (again) 2026-03-17 05:27:52,894 INFO common.Util: dfs.datanode.fileio.profiling.sampling.percentage set to 0. Disabling file IO profiling 2026-03-17 05:27:52,895 INFO datanode.BlockScanner: Initialized block scanner with targetBytesPerSec 1048576 2026-03-17 05:27:52,897 INFO datanode.DataNode: Configured hostname is 127.0.0.1 2026-03-17 05:27:52,898 INFO common.Util: dfs.datanode.fileio.profiling.sampling.percentage set to 0. Disabling file IO profiling 2026-03-17 05:27:52,900 INFO datanode.DataNode: Starting DataNode with maxLockedMemory = 0 2026-03-17 05:27:52,903 INFO datanode.DataNode: Opened streaming server at /127.0.0.1:35797 2026-03-17 05:27:52,904 INFO datanode.DataNode: Balancing bandwidth is 104857600 bytes/s 2026-03-17 05:27:52,904 INFO datanode.DataNode: Number threads for balancing is 100 2026-03-17 05:27:52,915 INFO server.AuthenticationFilter: Unable to initialize FileSignerSecretProvider, falling back to use random secrets. 2026-03-17 05:27:52,915 INFO http.HttpRequestLog: Http request log for http.requests.datanode is not defined 2026-03-17 05:27:52,916 INFO http.HttpServer2: Added global filter 'safety' (class=org.apache.hadoop.http.HttpServer2$QuotingInputFilter) 2026-03-17 05:27:52,918 INFO http.HttpServer2: Added filter static_user_filter (class=org.apache.hadoop.http.lib.StaticUserWebFilter$StaticUserFilter) to context datanode 2026-03-17 05:27:52,918 INFO http.HttpServer2: Added filter static_user_filter (class=org.apache.hadoop.http.lib.StaticUserWebFilter$StaticUserFilter) to context static 2026-03-17 05:27:52,918 INFO http.HttpServer2: Added filter static_user_filter (class=org.apache.hadoop.http.lib.StaticUserWebFilter$StaticUserFilter) to context logs 2026-03-17 05:27:52,919 INFO http.HttpServer2: Jetty bound to port 41777 2026-03-17 05:27:52,919 INFO server.Server: jetty-9.4.40.v20210413; built: 2021-04-13T20:42:42.668Z; git: b881a572662e1943a14ae12e7e1207989f218b74; jvm 11.0.25+9-post-Ubuntu-1ubuntu122.04 2026-03-17 05:27:52,920 INFO server.session: DefaultSessionIdManager workerName=node0 2026-03-17 05:27:52,920 INFO server.session: No SessionScavenger set, using defaults 2026-03-17 05:27:52,920 INFO server.session: node0 Scavenging every 660000ms 2026-03-17 05:27:52,921 INFO handler.ContextHandler: Started o.e.j.s.ServletContextHandler@5875de6a{logs,/logs,file:///hadoop-3.3.1/logs,AVAILABLE} 2026-03-17 05:27:52,921 INFO handler.ContextHandler: Started o.e.j.s.ServletContextHandler@2687725a{static,/static,file:///hadoop-3.3.1/share/hadoop/hdfs/webapps/static/,AVAILABLE} 2026-03-17 05:27:52,926 INFO handler.ContextHandler: Started o.e.j.w.WebAppContext@38e7ed69{datanode,/,file:///hadoop-3.3.1/share/hadoop/hdfs/webapps/datanode/,AVAILABLE}{file:/hadoop-3.3.1/share/hadoop/hdfs/webapps/datanode} 2026-03-17 05:27:52,927 INFO server.AbstractConnector: Started ServerConnector@40e32762{HTTP/1.1, (http/1.1)}{localhost:41777} 2026-03-17 05:27:52,927 INFO server.Server: Started @1871ms 2026-03-17 05:27:53,004 WARN web.DatanodeHttpServer: Got null for restCsrfPreventionFilter - will not do any filtering. 2026-03-17 05:27:53,054 INFO web.DatanodeHttpServer: Listening HTTP traffic on /127.0.0.1:33609 2026-03-17 05:27:53,055 INFO util.JvmPauseMonitor: Starting JVM pause monitor 2026-03-17 05:27:53,055 INFO datanode.DataNode: dnUserName = clickhouse 2026-03-17 05:27:53,056 INFO datanode.DataNode: supergroup = supergroup 2026-03-17 05:27:53,071 INFO ipc.CallQueueManager: Using callQueue: class java.util.concurrent.LinkedBlockingQueue, queueCapacity: 1000, scheduler: class org.apache.hadoop.ipc.DefaultRpcScheduler, ipcBackoff: false. 2026-03-17 05:27:53,072 INFO ipc.Server: Starting Socket Reader #1 for port 0 2026-03-17 05:27:53,079 INFO datanode.DataNode: Opened IPC server at /127.0.0.1:43949 2026-03-17 05:27:53,100 INFO datanode.DataNode: Refresh request received for nameservices: null 2026-03-17 05:27:53,100 INFO datanode.DataNode: Starting BPOfferServices for nameservices: 2026-03-17 05:27:53,106 INFO datanode.DataNode: Block pool (Datanode Uuid unassigned) service to localhost/127.0.0.1:12222 starting to offer service 2026-03-17 05:27:53,109 INFO ipc.Server: IPC Server Responder: starting 2026-03-17 05:27:53,111 INFO ipc.Server: IPC Server listener on 0: starting 2026-03-17 05:27:53,283 INFO datanode.DataNode: Acknowledging ACTIVE Namenode during handshakeBlock pool (Datanode Uuid unassigned) service to localhost/127.0.0.1:12222 2026-03-17 05:27:53,286 INFO common.Storage: Using 2 threads to upgrade data directories (dfs.datanode.parallel.volumes.load.threads.num=2, dataDirs=2) 2026-03-17 05:27:53,289 INFO common.Storage: Lock on /hadoop-3.3.1/target/test/data/dfs/data/data1/in_use.lock acquired by nodename 413@8de11e8011ac 2026-03-17 05:27:53,290 INFO common.Storage: Storage directory with location [DISK]file:/hadoop-3.3.1/target/test/data/dfs/data/data1 is not formatted for namespace 1227402346. Formatting... 2026-03-17 05:27:53,291 INFO common.Storage: Generated new storageID DS-5ea1d9ec-9648-48aa-bda3-5b075d872263 for directory /hadoop-3.3.1/target/test/data/dfs/data/data1 2026-03-17 05:27:53,297 INFO common.Storage: Lock on /hadoop-3.3.1/target/test/data/dfs/data/data2/in_use.lock acquired by nodename 413@8de11e8011ac 2026-03-17 05:27:53,298 INFO common.Storage: Storage directory with location [DISK]file:/hadoop-3.3.1/target/test/data/dfs/data/data2 is not formatted for namespace 1227402346. Formatting... 2026-03-17 05:27:53,298 INFO common.Storage: Generated new storageID DS-d0c56a90-49ab-446f-9673-980f2433c6b1 for directory /hadoop-3.3.1/target/test/data/dfs/data/data2 2026-03-17 05:27:53,326 INFO common.Storage: Analyzing storage directories for bpid BP-1367658773-172.17.0.2-1773761271981 2026-03-17 05:27:53,327 INFO common.Storage: Locking is disabled for /hadoop-3.3.1/target/test/data/dfs/data/data1/current/BP-1367658773-172.17.0.2-1773761271981 2026-03-17 05:27:53,328 INFO common.Storage: Block pool storage directory for location [DISK]file:/hadoop-3.3.1/target/test/data/dfs/data/data1 and block pool id BP-1367658773-172.17.0.2-1773761271981 is not formatted. Formatting ... 2026-03-17 05:27:53,328 INFO common.Storage: Formatting block pool BP-1367658773-172.17.0.2-1773761271981 directory /hadoop-3.3.1/target/test/data/dfs/data/data1/current/BP-1367658773-172.17.0.2-1773761271981/current 2026-03-17 05:27:53,353 INFO hdfs.MiniDFSCluster: dnInfo.length != numDataNodes 2026-03-17 05:27:53,353 INFO hdfs.MiniDFSCluster: Waiting for cluster to become active 2026-03-17 05:27:53,354 INFO common.Storage: Analyzing storage directories for bpid BP-1367658773-172.17.0.2-1773761271981 2026-03-17 05:27:53,354 INFO common.Storage: Locking is disabled for /hadoop-3.3.1/target/test/data/dfs/data/data2/current/BP-1367658773-172.17.0.2-1773761271981 2026-03-17 05:27:53,355 INFO common.Storage: Block pool storage directory for location [DISK]file:/hadoop-3.3.1/target/test/data/dfs/data/data2 and block pool id BP-1367658773-172.17.0.2-1773761271981 is not formatted. Formatting ... 2026-03-17 05:27:53,355 INFO common.Storage: Formatting block pool BP-1367658773-172.17.0.2-1773761271981 directory /hadoop-3.3.1/target/test/data/dfs/data/data2/current/BP-1367658773-172.17.0.2-1773761271981/current 2026-03-17 05:27:53,358 INFO datanode.DataNode: Setting up storage: nsid=1227402346;bpid=BP-1367658773-172.17.0.2-1773761271981;lv=-57;nsInfo=lv=-66;cid=testClusterID;nsid=1227402346;c=1773761271981;bpid=BP-1367658773-172.17.0.2-1773761271981;dnuuid=null 2026-03-17 05:27:53,362 INFO datanode.DataNode: Generated and persisted new Datanode UUID 63f05e85-b59f-4f1c-a004-c45689ed2170 2026-03-17 05:27:53,373 INFO impl.FsDatasetImpl: The datanode lock is a read write lock 2026-03-17 05:27:53,417 INFO impl.FsDatasetImpl: Added new volume: DS-5ea1d9ec-9648-48aa-bda3-5b075d872263 2026-03-17 05:27:53,418 INFO impl.FsDatasetImpl: Added volume - [DISK]file:/hadoop-3.3.1/target/test/data/dfs/data/data1, StorageType: DISK 2026-03-17 05:27:53,419 INFO impl.FsDatasetImpl: Added new volume: DS-d0c56a90-49ab-446f-9673-980f2433c6b1 2026-03-17 05:27:53,419 INFO impl.FsDatasetImpl: Added volume - [DISK]file:/hadoop-3.3.1/target/test/data/dfs/data/data2, StorageType: DISK 2026-03-17 05:27:53,422 INFO impl.MemoryMappableBlockLoader: Initializing cache loader: MemoryMappableBlockLoader. 2026-03-17 05:27:53,424 INFO impl.FsDatasetImpl: Registered FSDatasetState MBean 2026-03-17 05:27:53,428 INFO impl.FsDatasetImpl: Adding block pool BP-1367658773-172.17.0.2-1773761271981 2026-03-17 05:27:53,429 INFO impl.FsDatasetImpl: Scanning block pool BP-1367658773-172.17.0.2-1773761271981 on volume /hadoop-3.3.1/target/test/data/dfs/data/data2... 2026-03-17 05:27:53,428 INFO impl.FsDatasetImpl: Scanning block pool BP-1367658773-172.17.0.2-1773761271981 on volume /hadoop-3.3.1/target/test/data/dfs/data/data1... 2026-03-17 05:27:53,453 INFO impl.FsDatasetImpl: Time taken to scan block pool BP-1367658773-172.17.0.2-1773761271981 on /hadoop-3.3.1/target/test/data/dfs/data/data2: 24ms 2026-03-17 05:27:53,456 INFO hdfs.MiniDFSCluster: dnInfo.length != numDataNodes 2026-03-17 05:27:53,456 INFO hdfs.MiniDFSCluster: Waiting for cluster to become active 2026-03-17 05:27:53,465 INFO impl.FsDatasetImpl: Time taken to scan block pool BP-1367658773-172.17.0.2-1773761271981 on /hadoop-3.3.1/target/test/data/dfs/data/data1: 36ms 2026-03-17 05:27:53,465 INFO impl.FsDatasetImpl: Total time to scan all replicas for block pool BP-1367658773-172.17.0.2-1773761271981: 37ms 2026-03-17 05:27:53,466 INFO impl.FsDatasetImpl: Adding replicas to map for block pool BP-1367658773-172.17.0.2-1773761271981 on volume /hadoop-3.3.1/target/test/data/dfs/data/data1... 2026-03-17 05:27:53,466 INFO impl.FsDatasetImpl: Adding replicas to map for block pool BP-1367658773-172.17.0.2-1773761271981 on volume /hadoop-3.3.1/target/test/data/dfs/data/data2... 2026-03-17 05:27:53,466 INFO impl.BlockPoolSlice: Replica Cache file: /hadoop-3.3.1/target/test/data/dfs/data/data1/current/BP-1367658773-172.17.0.2-1773761271981/current/replicas doesn't exist 2026-03-17 05:27:53,466 INFO impl.BlockPoolSlice: Replica Cache file: /hadoop-3.3.1/target/test/data/dfs/data/data2/current/BP-1367658773-172.17.0.2-1773761271981/current/replicas doesn't exist 2026-03-17 05:27:53,468 INFO impl.FsDatasetImpl: Time to add replicas to map for block pool BP-1367658773-172.17.0.2-1773761271981 on volume /hadoop-3.3.1/target/test/data/dfs/data/data2: 2ms 2026-03-17 05:27:53,468 INFO impl.FsDatasetImpl: Time to add replicas to map for block pool BP-1367658773-172.17.0.2-1773761271981 on volume /hadoop-3.3.1/target/test/data/dfs/data/data1: 2ms 2026-03-17 05:27:53,468 INFO impl.FsDatasetImpl: Total time to add all replicas to map for block pool BP-1367658773-172.17.0.2-1773761271981: 3ms 2026-03-17 05:27:53,469 INFO checker.ThrottledAsyncChecker: Scheduling a check for /hadoop-3.3.1/target/test/data/dfs/data/data1 2026-03-17 05:27:53,475 INFO checker.DatasetVolumeChecker: Scheduled health check for volume /hadoop-3.3.1/target/test/data/dfs/data/data1 2026-03-17 05:27:53,475 INFO checker.ThrottledAsyncChecker: Scheduling a check for /hadoop-3.3.1/target/test/data/dfs/data/data2 2026-03-17 05:27:53,476 INFO checker.DatasetVolumeChecker: Scheduled health check for volume /hadoop-3.3.1/target/test/data/dfs/data/data2 2026-03-17 05:27:53,477 INFO datanode.VolumeScanner: Now scanning bpid BP-1367658773-172.17.0.2-1773761271981 on volume /hadoop-3.3.1/target/test/data/dfs/data/data1 2026-03-17 05:27:53,477 INFO datanode.VolumeScanner: Now scanning bpid BP-1367658773-172.17.0.2-1773761271981 on volume /hadoop-3.3.1/target/test/data/dfs/data/data2 2026-03-17 05:27:53,479 WARN datanode.DirectoryScanner: dfs.datanode.directoryscan.throttle.limit.ms.per.sec set to value above 1000 ms/sec. Assuming default value of -1 2026-03-17 05:27:53,479 INFO datanode.DirectoryScanner: Periodic Directory Tree Verification scan starting in 10848448ms with interval of 21600000ms and throttle limit of -1ms/s 2026-03-17 05:27:53,479 INFO datanode.VolumeScanner: VolumeScanner(/hadoop-3.3.1/target/test/data/dfs/data/data1, DS-5ea1d9ec-9648-48aa-bda3-5b075d872263): finished scanning block pool BP-1367658773-172.17.0.2-1773761271981 2026-03-17 05:27:53,479 INFO datanode.VolumeScanner: VolumeScanner(/hadoop-3.3.1/target/test/data/dfs/data/data2, DS-d0c56a90-49ab-446f-9673-980f2433c6b1): finished scanning block pool BP-1367658773-172.17.0.2-1773761271981 2026-03-17 05:27:53,484 INFO datanode.DataNode: Block pool BP-1367658773-172.17.0.2-1773761271981 (Datanode Uuid 63f05e85-b59f-4f1c-a004-c45689ed2170) service to localhost/127.0.0.1:12222 beginning handshake with NN 2026-03-17 05:27:53,494 INFO datanode.VolumeScanner: VolumeScanner(/hadoop-3.3.1/target/test/data/dfs/data/data2, DS-d0c56a90-49ab-446f-9673-980f2433c6b1): no suitable block pools found to scan. Waiting 1814399983 ms. 2026-03-17 05:27:53,494 INFO datanode.VolumeScanner: VolumeScanner(/hadoop-3.3.1/target/test/data/dfs/data/data1, DS-5ea1d9ec-9648-48aa-bda3-5b075d872263): no suitable block pools found to scan. Waiting 1814399983 ms. 2026-03-17 05:27:53,495 INFO hdfs.StateChange: BLOCK* registerDatanode: from DatanodeRegistration(127.0.0.1:35797, datanodeUuid=63f05e85-b59f-4f1c-a004-c45689ed2170, infoPort=33609, infoSecurePort=0, ipcPort=43949, storageInfo=lv=-57;cid=testClusterID;nsid=1227402346;c=1773761271981) storage 63f05e85-b59f-4f1c-a004-c45689ed2170 2026-03-17 05:27:53,496 INFO net.NetworkTopology: Adding a new node: /default-rack/127.0.0.1:35797 2026-03-17 05:27:53,496 INFO blockmanagement.BlockReportLeaseManager: Registered DN 63f05e85-b59f-4f1c-a004-c45689ed2170 (127.0.0.1:35797). 2026-03-17 05:27:53,500 INFO datanode.DataNode: Block pool BP-1367658773-172.17.0.2-1773761271981 (Datanode Uuid 63f05e85-b59f-4f1c-a004-c45689ed2170) service to localhost/127.0.0.1:12222 successfully registered with NN 2026-03-17 05:27:53,500 INFO datanode.DataNode: For namenode localhost/127.0.0.1:12222 using BLOCKREPORT_INTERVAL of 21600000msecs CACHEREPORT_INTERVAL of 10000msecs Initial delay: 0msecs; heartBeatInterval=3000 2026-03-17 05:27:53,520 INFO blockmanagement.DatanodeDescriptor: Adding new storage ID DS-5ea1d9ec-9648-48aa-bda3-5b075d872263 for DN 127.0.0.1:35797 2026-03-17 05:27:53,521 INFO blockmanagement.DatanodeDescriptor: Adding new storage ID DS-d0c56a90-49ab-446f-9673-980f2433c6b1 for DN 127.0.0.1:35797 2026-03-17 05:27:53,553 INFO BlockStateChange: BLOCK* processReport 0x23598960236aa008: Processing first storage report for DS-5ea1d9ec-9648-48aa-bda3-5b075d872263 from datanode 63f05e85-b59f-4f1c-a004-c45689ed2170 2026-03-17 05:27:53,555 INFO BlockStateChange: BLOCK* processReport 0x23598960236aa008: from storage DS-5ea1d9ec-9648-48aa-bda3-5b075d872263 node DatanodeRegistration(127.0.0.1:35797, datanodeUuid=63f05e85-b59f-4f1c-a004-c45689ed2170, infoPort=33609, infoSecurePort=0, ipcPort=43949, storageInfo=lv=-57;cid=testClusterID;nsid=1227402346;c=1773761271981), blocks: 0, hasStaleStorage: true, processing time: 2 msecs, invalidatedBlocks: 0 2026-03-17 05:27:53,555 INFO BlockStateChange: BLOCK* processReport 0x23598960236aa008: Processing first storage report for DS-d0c56a90-49ab-446f-9673-980f2433c6b1 from datanode 63f05e85-b59f-4f1c-a004-c45689ed2170 2026-03-17 05:27:53,556 INFO BlockStateChange: BLOCK* processReport 0x23598960236aa008: from storage DS-d0c56a90-49ab-446f-9673-980f2433c6b1 node DatanodeRegistration(127.0.0.1:35797, datanodeUuid=63f05e85-b59f-4f1c-a004-c45689ed2170, infoPort=33609, infoSecurePort=0, ipcPort=43949, storageInfo=lv=-57;cid=testClusterID;nsid=1227402346;c=1773761271981), blocks: 0, hasStaleStorage: false, processing time: 0 msecs, invalidatedBlocks: 0 2026-03-17 05:27:53,565 INFO hdfs.MiniDFSCluster: Cluster is active 2026-03-17 05:27:53,570 INFO mapreduce.MiniHadoopClusterManager: Started MiniDFSCluster -- namenode on port 12222 2026-03-17 05:27:53,576 INFO datanode.DataNode: Successfully sent block report 0x23598960236aa008 to namenode: localhost/127.0.0.1:12222, containing 2 storage report(s), of which we sent 2. The reports had 0 total blocks and used 1 RPC(s). This took 5 msecs to generate and 37 msecs for RPC and NN processing. Got back one command: FinalizeCommand/5. 2026-03-17 05:27:53,577 INFO datanode.DataNode: Got finalize command for block pool BP-1367658773-172.17.0.2-1773761271981 2026-03-17 05:38:21,107 INFO namenode.FSEditLog: Number of transactions: 2 Total time for transactions(ms): 53 Number of transactions batched in Syncs: 0 Number of syncs: 2 SyncTimes(ms): 1 1 2026-03-17 05:38:21,131 INFO hdfs.StateChange: BLOCK* allocate blk_1073741825_1001, replicas=127.0.0.1:35797 for /02923_hdfs_engine_size_virtual_column_test_dqxg6ci1.data1.tsv 2026-03-17 05:38:21,259 INFO datanode.DataNode: Receiving BP-1367658773-172.17.0.2-1773761271981:blk_1073741825_1001 src: /127.0.0.1:36416 dest: /127.0.0.1:35797 2026-03-17 05:38:21,286 INFO DataNode.clienttrace: src: /127.0.0.1:36416, dest: /127.0.0.1:35797, bytes: 2, op: HDFS_WRITE, cliID: libhdfs3_client_rand_0.639439_count_2_pid_599_tid_140069572486720, offset: 0, srvID: 63f05e85-b59f-4f1c-a004-c45689ed2170, blockid: BP-1367658773-172.17.0.2-1773761271981:blk_1073741825_1001, duration(ns): 6730870 2026-03-17 05:38:21,286 INFO datanode.DataNode: PacketResponder: BP-1367658773-172.17.0.2-1773761271981:blk_1073741825_1001, type=LAST_IN_PIPELINE terminating 2026-03-17 05:38:21,289 INFO hdfs.StateChange: BLOCK* fsync: /02923_hdfs_engine_size_virtual_column_test_dqxg6ci1.data1.tsv for libhdfs3_client_rand_0.639439_count_2_pid_599_tid_140069572486720 2026-03-17 05:38:21,294 INFO namenode.FSNamesystem: BLOCK* blk_1073741825_1001 is COMMITTED but not COMPLETE(numNodes= 0 < minimum = 1) in file /02923_hdfs_engine_size_virtual_column_test_dqxg6ci1.data1.tsv 2026-03-17 05:38:21,698 INFO hdfs.StateChange: DIR* completeFile: /02923_hdfs_engine_size_virtual_column_test_dqxg6ci1.data1.tsv is closed by libhdfs3_client_rand_0.639439_count_2_pid_599_tid_140069572486720 2026-03-17 05:38:22,105 INFO hdfs.StateChange: BLOCK* allocate blk_1073741826_1002, replicas=127.0.0.1:35797 for /02923_hdfs_engine_size_virtual_column_test_dqxg6ci1.data2.tsv 2026-03-17 05:38:22,108 INFO datanode.DataNode: Receiving BP-1367658773-172.17.0.2-1773761271981:blk_1073741826_1002 src: /127.0.0.1:35710 dest: /127.0.0.1:35797 2026-03-17 05:38:22,112 INFO DataNode.clienttrace: src: /127.0.0.1:35710, dest: /127.0.0.1:35797, bytes: 3, op: HDFS_WRITE, cliID: libhdfs3_client_rand_0.724472_count_4_pid_599_tid_140077690684992, offset: 0, srvID: 63f05e85-b59f-4f1c-a004-c45689ed2170, blockid: BP-1367658773-172.17.0.2-1773761271981:blk_1073741826_1002, duration(ns): 1945053 2026-03-17 05:38:22,113 INFO datanode.DataNode: PacketResponder: BP-1367658773-172.17.0.2-1773761271981:blk_1073741826_1002, type=LAST_IN_PIPELINE terminating 2026-03-17 05:38:22,113 INFO hdfs.StateChange: BLOCK* fsync: /02923_hdfs_engine_size_virtual_column_test_dqxg6ci1.data2.tsv for libhdfs3_client_rand_0.724472_count_4_pid_599_tid_140077690684992 2026-03-17 05:38:22,115 INFO hdfs.StateChange: DIR* completeFile: /02923_hdfs_engine_size_virtual_column_test_dqxg6ci1.data2.tsv is closed by libhdfs3_client_rand_0.724472_count_4_pid_599_tid_140077690684992 2026-03-17 05:38:22,564 INFO hdfs.StateChange: BLOCK* allocate blk_1073741827_1003, replicas=127.0.0.1:35797 for /02923_hdfs_engine_size_virtual_column_test_dqxg6ci1.data3.tsv 2026-03-17 05:38:22,566 INFO datanode.DataNode: Receiving BP-1367658773-172.17.0.2-1773761271981:blk_1073741827_1003 src: /127.0.0.1:35718 dest: /127.0.0.1:35797 2026-03-17 05:38:22,569 INFO DataNode.clienttrace: src: /127.0.0.1:35718, dest: /127.0.0.1:35797, bytes: 4, op: HDFS_WRITE, cliID: libhdfs3_client_rand_0.724472_count_6_pid_599_tid_140069619238464, offset: 0, srvID: 63f05e85-b59f-4f1c-a004-c45689ed2170, blockid: BP-1367658773-172.17.0.2-1773761271981:blk_1073741827_1003, duration(ns): 1612199 2026-03-17 05:38:22,569 INFO datanode.DataNode: PacketResponder: BP-1367658773-172.17.0.2-1773761271981:blk_1073741827_1003, type=LAST_IN_PIPELINE terminating 2026-03-17 05:38:22,570 INFO hdfs.StateChange: BLOCK* fsync: /02923_hdfs_engine_size_virtual_column_test_dqxg6ci1.data3.tsv for libhdfs3_client_rand_0.724472_count_6_pid_599_tid_140069619238464 2026-03-17 05:38:22,571 INFO hdfs.StateChange: DIR* completeFile: /02923_hdfs_engine_size_virtual_column_test_dqxg6ci1.data3.tsv is closed by libhdfs3_client_rand_0.724472_count_6_pid_599_tid_140069619238464 2026-03-17 05:54:35,769 INFO namenode.FSEditLog: Number of transactions: 20 Total time for transactions(ms): 56 Number of transactions batched in Syncs: 4 Number of syncs: 16 SyncTimes(ms): 1 3 2026-03-17 05:54:35,784 INFO hdfs.StateChange: BLOCK* allocate blk_1073741828_1004, replicas=127.0.0.1:35797 for /test_02725_1.tsv 2026-03-17 05:54:35,795 INFO datanode.DataNode: Receiving BP-1367658773-172.17.0.2-1773761271981:blk_1073741828_1004 src: /127.0.0.1:40220 dest: /127.0.0.1:35797 2026-03-17 05:54:35,802 INFO DataNode.clienttrace: src: /127.0.0.1:40220, dest: /127.0.0.1:35797, bytes: 6, op: HDFS_WRITE, cliID: libhdfs3_client_rand_0.461045_count_16_pid_599_tid_140042208704064, offset: 0, srvID: 63f05e85-b59f-4f1c-a004-c45689ed2170, blockid: BP-1367658773-172.17.0.2-1773761271981:blk_1073741828_1004, duration(ns): 3359088 2026-03-17 05:54:35,803 INFO datanode.DataNode: PacketResponder: BP-1367658773-172.17.0.2-1773761271981:blk_1073741828_1004, type=LAST_IN_PIPELINE terminating 2026-03-17 05:54:35,804 INFO hdfs.StateChange: BLOCK* fsync: /test_02725_1.tsv for libhdfs3_client_rand_0.461045_count_16_pid_599_tid_140042208704064 2026-03-17 05:54:35,807 INFO hdfs.StateChange: DIR* completeFile: /test_02725_1.tsv is closed by libhdfs3_client_rand_0.461045_count_16_pid_599_tid_140042208704064 2026-03-17 05:54:36,234 INFO hdfs.StateChange: BLOCK* allocate blk_1073741829_1005, replicas=127.0.0.1:35797 for /test_02725_2.tsv 2026-03-17 05:54:36,237 INFO datanode.DataNode: Receiving BP-1367658773-172.17.0.2-1773761271981:blk_1073741829_1005 src: /127.0.0.1:40222 dest: /127.0.0.1:35797 2026-03-17 05:54:36,241 INFO DataNode.clienttrace: src: /127.0.0.1:40222, dest: /127.0.0.1:35797, bytes: 6, op: HDFS_WRITE, cliID: libhdfs3_client_rand_0.546077_count_18_pid_599_tid_140042208704064, offset: 0, srvID: 63f05e85-b59f-4f1c-a004-c45689ed2170, blockid: BP-1367658773-172.17.0.2-1773761271981:blk_1073741829_1005, duration(ns): 2540694 2026-03-17 05:54:36,242 INFO datanode.DataNode: PacketResponder: BP-1367658773-172.17.0.2-1773761271981:blk_1073741829_1005, type=LAST_IN_PIPELINE terminating 2026-03-17 05:54:36,242 INFO hdfs.StateChange: BLOCK* fsync: /test_02725_2.tsv for libhdfs3_client_rand_0.546077_count_18_pid_599_tid_140042208704064 2026-03-17 05:54:36,244 INFO hdfs.StateChange: DIR* completeFile: /test_02725_2.tsv is closed by libhdfs3_client_rand_0.546077_count_18_pid_599_tid_140042208704064 2026-03-17 05:55:37,421 INFO namenode.FSEditLog: Number of transactions: 32 Total time for transactions(ms): 58 Number of transactions batched in Syncs: 6 Number of syncs: 26 SyncTimes(ms): 3 4 2026-03-17 05:55:37,429 INFO hdfs.StateChange: BLOCK* allocate blk_1073741830_1006, replicas=127.0.0.1:35797 for /test_02536.jsonl 2026-03-17 05:55:37,432 INFO datanode.DataNode: Receiving BP-1367658773-172.17.0.2-1773761271981:blk_1073741830_1006 src: /127.0.0.1:36886 dest: /127.0.0.1:35797 2026-03-17 05:55:37,435 INFO DataNode.clienttrace: src: /127.0.0.1:36886, dest: /127.0.0.1:35797, bytes: 27, op: HDFS_WRITE, cliID: libhdfs3_client_rand_0.733057_count_27_pid_599_tid_140069401204288, offset: 0, srvID: 63f05e85-b59f-4f1c-a004-c45689ed2170, blockid: BP-1367658773-172.17.0.2-1773761271981:blk_1073741830_1006, duration(ns): 1524487 2026-03-17 05:55:37,436 INFO datanode.DataNode: PacketResponder: BP-1367658773-172.17.0.2-1773761271981:blk_1073741830_1006, type=LAST_IN_PIPELINE terminating 2026-03-17 05:55:37,436 INFO hdfs.StateChange: BLOCK* fsync: /test_02536.jsonl for libhdfs3_client_rand_0.733057_count_27_pid_599_tid_140069401204288 2026-03-17 05:55:37,438 INFO hdfs.StateChange: DIR* completeFile: /test_02536.jsonl is closed by libhdfs3_client_rand_0.733057_count_27_pid_599_tid_140069401204288 2026-03-17 06:02:45,796 INFO namenode.FSEditLog: Number of transactions: 38 Total time for transactions(ms): 58 Number of transactions batched in Syncs: 7 Number of syncs: 31 SyncTimes(ms): 3 5 2026-03-17 06:02:45,804 INFO hdfs.StateChange: BLOCK* allocate blk_1073741831_1007, replicas=127.0.0.1:35797 for /test_1.tsv 2026-03-17 06:02:45,807 INFO datanode.DataNode: Receiving BP-1367658773-172.17.0.2-1773761271981:blk_1073741831_1007 src: /127.0.0.1:45136 dest: /127.0.0.1:35797 2026-03-17 06:02:45,812 INFO DataNode.clienttrace: src: /127.0.0.1:45136, dest: /127.0.0.1:35797, bytes: 6, op: HDFS_WRITE, cliID: libhdfs3_client_rand_0.126945_count_34_pid_599_tid_140016079189568, offset: 0, srvID: 63f05e85-b59f-4f1c-a004-c45689ed2170, blockid: BP-1367658773-172.17.0.2-1773761271981:blk_1073741831_1007, duration(ns): 1998398 2026-03-17 06:02:45,812 INFO datanode.DataNode: PacketResponder: BP-1367658773-172.17.0.2-1773761271981:blk_1073741831_1007, type=LAST_IN_PIPELINE terminating 2026-03-17 06:02:45,813 INFO hdfs.StateChange: BLOCK* fsync: /test_1.tsv for libhdfs3_client_rand_0.126945_count_34_pid_599_tid_140016079189568 2026-03-17 06:02:45,814 INFO hdfs.StateChange: DIR* completeFile: /test_1.tsv is closed by libhdfs3_client_rand_0.126945_count_34_pid_599_tid_140016079189568 2026-03-17 06:02:45,871 INFO hdfs.StateChange: BLOCK* allocate blk_1073741832_1008, replicas=127.0.0.1:35797 for /test_2.tsv 2026-03-17 06:02:45,873 INFO datanode.DataNode: Receiving BP-1367658773-172.17.0.2-1773761271981:blk_1073741832_1008 src: /127.0.0.1:45150 dest: /127.0.0.1:35797 2026-03-17 06:02:45,877 INFO DataNode.clienttrace: src: /127.0.0.1:45150, dest: /127.0.0.1:35797, bytes: 6, op: HDFS_WRITE, cliID: libhdfs3_client_rand_0.126945_count_36_pid_599_tid_140016079189568, offset: 0, srvID: 63f05e85-b59f-4f1c-a004-c45689ed2170, blockid: BP-1367658773-172.17.0.2-1773761271981:blk_1073741832_1008, duration(ns): 1735334 2026-03-17 06:02:45,877 INFO datanode.DataNode: PacketResponder: BP-1367658773-172.17.0.2-1773761271981:blk_1073741832_1008, type=LAST_IN_PIPELINE terminating 2026-03-17 06:02:45,879 INFO hdfs.StateChange: BLOCK* fsync: /test_2.tsv for libhdfs3_client_rand_0.126945_count_36_pid_599_tid_140016079189568 2026-03-17 06:02:45,880 INFO hdfs.StateChange: DIR* completeFile: /test_2.tsv is closed by libhdfs3_client_rand_0.126945_count_36_pid_599_tid_140016079189568 2026-03-17 06:02:45,911 INFO hdfs.StateChange: BLOCK* allocate blk_1073741833_1009, replicas=127.0.0.1:35797 for /test_3.tsv 2026-03-17 06:02:45,914 INFO datanode.DataNode: Receiving BP-1367658773-172.17.0.2-1773761271981:blk_1073741833_1009 src: /127.0.0.1:45162 dest: /127.0.0.1:35797 2026-03-17 06:02:45,918 INFO DataNode.clienttrace: src: /127.0.0.1:45162, dest: /127.0.0.1:35797, bytes: 6, op: HDFS_WRITE, cliID: libhdfs3_client_rand_0.126945_count_38_pid_599_tid_140016079189568, offset: 0, srvID: 63f05e85-b59f-4f1c-a004-c45689ed2170, blockid: BP-1367658773-172.17.0.2-1773761271981:blk_1073741833_1009, duration(ns): 2339499 2026-03-17 06:02:45,918 INFO datanode.DataNode: PacketResponder: BP-1367658773-172.17.0.2-1773761271981:blk_1073741833_1009, type=LAST_IN_PIPELINE terminating 2026-03-17 06:02:45,919 INFO hdfs.StateChange: BLOCK* fsync: /test_3.tsv for libhdfs3_client_rand_0.126945_count_38_pid_599_tid_140016079189568 2026-03-17 06:02:45,920 INFO hdfs.StateChange: DIR* completeFile: /test_3.tsv is closed by libhdfs3_client_rand_0.126945_count_38_pid_599_tid_140016079189568